making process for the data shown in Figure 3.37(b) can be
ated in Figure 3.38(b). These trees certainly help visualise how
decisions based on the optimised partitioning rules. For instance,
sion-making process shown in Figure 3.38(b) can be written as
hich is very similar to the human intelligence,
y is less than two and x is negative, the object is a triangle.
kind of human-intelligence-alike models may not be easily
using the aforementioned machine learning algorithms such as
MLP for the time being.
he purity measurements
on-making model is a system composed of a set of partitioning
derive an optimal set of partitioning rules as the best decision-
model for a data set is an inductive learning process. For instance,
s certainly not a good partitioning rule compared with the
ng rule y = 2 in Figure 3.37(a). This is because the upper subspace
d by this partitioning rule is highly mixed by two classes of data
hough the lower subspace is very pure to one class of data points.
her hand, each subspace generated by the partitioning rule y = 2
ure to one class of data points.
utomate the discovery of an optimal partitioning rule set, a
ve measurement for measuring the purity or impurity of the
s generated by every partitioning rule is required. Two
ments have been proposed and are still in use. One is called the
x used by CART [Breiman, et al., 1984] and the other is called
mation gain used by both DT and CART [Breiman, et al., 1984;
1986].
Gini index is calculated for each portioning rule. It is the class-
of the products between the probability of correct classification
robability of mis-classification. The product ሺ߬ሻሾ1 െሺ߬ሻሿ is